AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Single GPU Training

# Single GPU Training

Babyberta 3
MIT
BabyBERTa is a lightweight version based on RoBERTa, specifically designed for language acquisition research, trained on a 5-million-word corpus of American English child-directed input.
Large Language Model Transformers English
B
phueb
22
0
Chinese Mobile Bert
Apache-2.0
This model was pre-trained on a 250-million-word Chinese corpus using the MobileBERT architecture, with a training period of 15 days, completing 1 million iterations on a single A100 GPU.
Large Language Model Transformers
C
Ayou
25
5
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase